posted 10-04-2012 03:00 PM
The meta-analysis was conducted with the assumption (as Cleve Backster would say) of ideally formulated relevant questions. That is, we made no attempt to differentiate or study the effectiveness of different types of questions. Instead, we assumed that questions were (mostly) adequate to prompt the examinee to respond to due to some combination of emotion, cognition, and behaviorally conditioned experience regarding the target stimuli. We could also assume there is a normal degree of imperfection in the questions used in the included studies. The bottom line is this: ain't nothin' perfect in life. If we do nothing until we can control all variable then we (and everyeone else) would do... nothing. So, we're not done, and there is obviously more to learn later.
Research can be conducted at a variety of levels. One of which would be to investigate the effectiveness of different types of exams (ie. PCSOT, law-enforcement applicant, information security, criminal investigation, etc.)
The hypothesis would be that different types of topics or targets would produce different rates of accuracy. It is an interesting hypothesis - one that should someday be studied. However, studying this will require that someone articulate a credible hypothesis why different topics or target questions would produce different results. In the past we have engaged in all kinds of amateur psychologizing (read: mind-reading) - suggesting that some questions are too hot, or too emotional (which suggests that the polygraph measures temperature or emotion).
What the premise really seems to say is that examinees may respond too much to the stimuli. This, of course, begs the question: how much is too much response? And how much response is enough response? The real problem is that to answer these questions we would have to know in advance of the test how much we want someone to respond - we would essentially have to know in advance whether they were guilty or innocent.
It will be better to simply present the stimulus questions according to reasonable rules and principles based on what we actually know about psychology and physiological responding. Without any mind-reading.
There is ample evidence that the polygraph does not measure either temperature or emotion - and the emerging base of scientific evidence that we do have seems not to support the notion that different topics/questions perform very differently. Now, certainly there will be differences somewhere, but they may or may not be related to the domains of law-enforcement applicant, PCSOT, information security, or criminal investigation. Differences in question effectiveness may have more to do with general principles that are applicable to all domains of testing.
One example: in 2007 we showed a poster presentation on OSS-3 at the APA conference. OSS-3 results were shown with three samples of screening exams, for which the sampling methodology was that same - suboptimal criterion established by a deceptive test result coupled with a confession to a test question, or truthful test result that was supporte by two different QC examiners. One LEPET sample. One PCSOT maintenance sample. One PCSOT sex history sample. Resulting accuracy: essentially the same for all three samples.
The actual studies in the meta-analysis are too many for me to remember, but they are described in detail in the complete report.
r
------------------
"Gentlemen, you can't fight in here. This is the war room."
--(Stanley Kubrick/Peter Sellers - Dr. Strangelove, 1964)